# Large-scale parameter model

Nllb 200 3.3B Ctranslate2
NLLB-200 is a neural machine translation model supporting 200 languages, focusing on translation research for low-resource languages.
Machine Translation Supports Multiple Languages
N
entai2965
25
2
Nllb Moe 54b 4bit
NLLB-MoE is a Mixture of Experts machine translation model developed by Meta, supporting 200 languages, and is one of the most advanced open-access machine translation models available.
Machine Translation Transformers Supports Multiple Languages
N
KnutJaegersberg
17
5
Ernie 3.0 Xbase Zh
ERNIE 3.0 is a large-scale knowledge-enhanced pre-trained model for language understanding and generation, developed by Baidu.
Large Language Model Transformers Chinese
E
nghuyong
14.27k
20
Codet5 Large Ntp Py
Bsd-3-clause
CodeT5 is a large-scale encoder-decoder model pre-trained with NTP objectives for Python language, focusing on code understanding and generation tasks
Large Language Model Transformers
C
Salesforce
217
27
Blenderbot 1B Distill
Apache-2.0
This model is a high-performance open-domain chatbot capable of integrating multiple dialogue skills such as questioning, answering, knowledge demonstration, and empathy.
Dialogue System Transformers English
B
facebook
2,413
37
Blenderbot 400M Distill
Apache-2.0
This model achieves multi-skill integrated open-domain dialogue capabilities through large-scale neural models and carefully designed training strategies.
Dialogue System English
B
facebook
203.20k
431
Blenderbot Small 90M
Apache-2.0
This is an open-domain dialogue system based on large-scale neural networks, capable of conducting multi-turn natural conversations and integrating various dialogue skills.
Dialogue System English
B
facebook
2,407
49
T0pp
Apache-2.0
T0pp is an 11-billion-parameter encoder-decoder model based on the T5 architecture, excelling in zero-shot task generalization with English natural language prompts, outperforming GPT-3 while being more compact.
Large Language Model Transformers English
T
bigscience
7,426
401
Deberta V3 Large
MIT
DeBERTaV3 improves upon DeBERTa with ELECTRA-style pre-training and gradient-disentangled embedding sharing techniques, excelling in natural language understanding tasks
Large Language Model Transformers English
D
microsoft
343.39k
213
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase